An optimal subgradient algorithm for large-scale bound-constrained convex optimization

نویسندگان
چکیده

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

An optimal subgradient algorithm for large-scale bound-constrained convex optimization

This paper shows that the OSGA algorithm – which uses first-order information to solve convex optimization problems with optimal complexity – can be used to efficiently solve arbitrary bound-constrained convex optimization problems. This is done by constructing an explicit method as well as an inexact scheme for solving the bound-constrained rational subproblem required by OSGA. This leads to a...

متن کامل

An Optimal Subgradient Algorithm for Large-scale Convex Optimization in Simple Domains

This paper shows that the optimal subgradient algorithm, OSGA, proposed in [59] can be used for solving structured large-scale convex constrained optimization problems. Only firstorder information is required, and the optimal complexity bounds for both smooth and nonsmooth problems are attained. More specifically, we consider two classes of problems: (i) a convex objective with a simple closed ...

متن کامل

CONSTRAINED BIG BANG-BIG CRUNCH ALGORITHM FOR OPTIMAL SOLUTION OF LARGE SCALE RESERVOIR OPERATION PROBLEM

A constrained version of the Big Bang-Big Crunch algorithm for the efficient solution of the optimal reservoir operation problems is proposed in this paper. Big Bang-Big Crunch (BB-BC) algorithm is a new meta-heuristic population-based algorithm that relies on one of the theories of the evolution of universe namely, the Big Bang and Big Crunch theory. An improved formulation of the algorithm na...

متن کامل

An Optimal Algorithm for Constrained Differentiable Convex Optimization

We describe three algorithms for solving differentiable convex optimization problems constrained to simple sets in Rn, i.e., sets on which it is easy to project an arbitrary point. The first two algorithms are optimal in the sense that they achieve an absolute precision of ε in relation to the optimal value in O(1/ √ ε) iterations using only first order information. This complexity depends on a...

متن کامل

SNOPT : An SQP Algorithm for Large - Scale Constrained Optimization ∗ Philip

Sequential quadratic programming (SQP) methods have proved highly effective for solving constrained optimization problems with smooth nonlinear functions in the objective and constraints. Here we consider problems with general inequality constraints (linear and nonlinear). We assume that first derivatives are available and that the constraint gradients are sparse. Second derivatives are assumed...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Mathematical Methods of Operations Research

سال: 2017

ISSN: 1432-2994,1432-5217

DOI: 10.1007/s00186-017-0585-1